json-ext

A set of utilities that extend the use of JSON. Designed to be fast and memory efficient
Features:
Install
npm install @discoveryjs/json-ext
API
parseChunked(chunkEmitter)
Works the same as JSON.parse()
but takes chunkEmitter
instead of string and returns Promise.
NOTE: reviver
parameter is not supported yet, but will be added in next releases.
NOTE: WHATWG streams aren't supported yet
When to use:
- It's required to avoid freezing the main thread during big JSON parsing, since this process can be distributed in time
- Huge JSON needs to be parsed (e.g. >500MB on Node.js)
- Needed to reduce memory pressure.
JSON.parse()
needs to receive the entire JSON before parsing it. With parseChunked()
you may parse JSON as first bytes of it comes. This approach helps to avoid storing a huge string in the memory at a single time point and following GC.
Benchmark
Usage:
const { parseChunked } = require('@discoveryjs/json-ext');
parseChunked(chunkEmitter)
.then(data => {
});
const data = await parseChunked(chunkEmitter);
Parameter chunkEmitter
can be:
const fs = require('fs');
const { parseChunked } = require('@discoveryjs/json-ext');
parseChunked(fs.createReadStream('path/to/file.json'))
- Generator, async generator or function that returns iterable (chunks). Chunk might be a
string
, Uint8Array
or Buffer
(Node.js only):
const { parseChunked } = require('@discoveryjs/json-ext');
const encoder = new TextEncoder();
parseChunked(function*() {
yield '{ "hello":';
yield Buffer.from(' "wor');
yield encoder.encode('ld" }');
});
parseChunked(async function*() {
for await (const chunk of someAsyncSource) {
yield chunk;
}
});
parseChunked(() => ['{ "hello":', ' "world"}'])
Using with fetch():
async function loadData(url) {
const response = await fetch(url);
const reader = response.body.getReader();
return parseChunked(async function*() {
while (true) {
const { done, value } = await reader.read();
if (done) {
break;
}
yield value;
}
});
}
loadData('https://example.com/data.json')
.then(data => {
})
stringifyStream(value[, replacer[, space]])
Works the same as JSON.stringify()
, but returns an instance of ReadableStream
instead of string.
NOTE: WHATWG Streams aren't supported yet, so function available for Node.js only for now
Departs from JSON.stringify():
- Outputs
null
when JSON.stringify()
returns undefined
(since streams may not emit undefined
) - A promise is resolving and the resulting value is stringifying as a regular one
- A stream in non-object mode is piping to output as is
- A stream in object mode is piping to output as an array of objects
When to use:
- Huge JSON needs to be generated (e.g. >500MB on Node.js)
- Needed to reduce memory pressure.
JSON.stringify()
needs to generate the entire JSON before send or write it to somewhere. With stringifyStream()
you may send a result to somewhere as first bytes of the result appears. This approach helps to avoid storing a huge string in the memory at a single time point. - The object being serialized contains Promises or Streams (see Usage for examples)
Benchmark
Usage:
const { stringifyStream } = require('@discoveryjs/json-ext');
stringifyStream(data)
.on('data', chunk => console.log(chunk))
.on('error', error => consold.error(error))
.on('finish', () => console.log('DONE!'));
stringifyStream(data)
.pipe(writableStream);
Using Promise or ReadableStream in serializing object:
const fs = require('fs');
const { stringifyStream } = require('@discoveryjs/json-ext');
stringifyStream({
name: 'example',
willSerializeResolvedValue: Promise.resolve(42),
fromFile: fs.createReadStream('path/to/file.json'),
at: {
any: {
level: new Promise(resolve => setTimeout(() => resolve('promise!'), 100))
}
}
})
stringifyStream({
foo: fetch('http://example.com/request_takes_2s').then(req => req.json()),
bar: fetch('http://example.com/request_takes_5s').then(req => req.json())
});
Using with WritableStream
(Node.js only):
const fs = require('fs');
const { stringifyStream } = require('@discoveryjs/json-ext');
stringifyStream(data)
.pipe(process.stdout);
stringifyStream(data)
.pipe(fs.createWriteStream('path/to/file.json'));
new Promise((resolve, reject) => {
stringifyStream(data)
.on('error', reject)
.pipe(stream)
.on('error', reject)
.on('finish', resolve);
});
stringifyInfo(value[, replacer[, space[, options]]])
value
, replacer
and space
arguments are the same as for JSON.stringify()
.
Result is an object:
{
minLength: Number,
circular: [...],
duplicate: [...],
async: [...]
}
Example:
const { stringifyInfo } = require('@discoveryjs/json-ext');
console.log(
stringifyInfo({ test: true }).minLength
);
Options
async
Type: Boolean
Default: false
Collect async values (promises and streams) or not.
continueOnCircular
Type: Boolean
Default: false
Stop collecting info for a value or not whenever circular reference is found. Setting option to true
allows to find all circular references.
version
The version of library, e.g. "0.3.1"
.
License
MIT